A measure of relative entropy between individual sequences with application to universal classification

نویسندگان

  • Jacob Ziv
  • Neri Merhav
چکیده

A new notion of empirical informational divergence (relative entropy) between two individual sequences is introduced. If the two sequences are independent realizations of two finiteorder, finite alphabet, stationary Markov processes, the empirical relative entropy converges to the relative entropy almost surely. This new empirical divergence is based on a version of the Lempel-Ziv data compression algorithm. A simple universal classification algorithm for individual sequences into a finite number of classes which is based on the empirical divergence, is introduced. It discriminates between the classes whenever they are distinguishable by some finite-memory classifier, for almost every given training sets and almost any test sequence from these classes. It is universal in the sense of being independent of the unknown sources. Index Tem-Lempel-Ziv algorithm, information divergence, finite-membry machines, finite-state machines, universal classification, universal hypothesis testing.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

A research on classification performance of fuzzy classifiers based on fuzzy set theory

Due to the complexities of objects and the vagueness of the human mind, it has attracted considerable attention from researchers studying fuzzy classification algorithms. In this paper, we propose a concept of fuzzy relative entropy to measure the divergence between two fuzzy sets. Applying fuzzy relative entropy, we prove the conclusion that patterns with high fuzziness are close to the classi...

متن کامل

On the optimality of universal classifiers for finite-length individual test sequences

On the optimality of universal classifiers for finite-length individual test sequences. Abstract It has been demonstrated that if two individual sequences are independent realizations of two finite-order, finite alphabet, stationary Markov processes, an empirical divergence measure (ZMM) that is based on cross-parsing of one sequence relative to the second one converges to the relative entropy ...

متن کامل

Multiplication Free Holographic Coding

An empirical informational divergence (relative entropy) between two individual sequences has been introduced in [1]. It has been demonstrated that if the two sequences are independent realizations of two finite-order, finite alphabet, stationary Markov processes, the proposed empirical divergence measure (ZMM), converges to the relative entropy almost surely. This leads to a realization of an ...

متن کامل

Modification of the Fast Global K-means Using a Fuzzy Relation with Application in Microarray Data Analysis

Recognizing genes with distinctive expression levels can help in prevention, diagnosis and treatment of the diseases at the genomic level. In this paper, fast Global k-means (fast GKM) is developed for clustering the gene expression datasets. Fast GKM is a significant improvement of the k-means clustering method. It is an incremental clustering method which starts with one cluster. Iteratively ...

متن کامل

Observational Modeling of the Kolmogorov-Sinai Entropy

In this paper, Kolmogorov-Sinai entropy is studied using mathematical modeling of an observer $ Theta $. The relative entropy of a sub-$ sigma_Theta $-algebra having finite atoms is defined and then   the ergodic properties of relative  semi-dynamical systems are investigated.  Also,  a relative version of Kolmogorov-Sinai theorem  is given. Finally, it is proved  that the relative entropy of a...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:
  • IEEE Trans. Information Theory

دوره 39  شماره 

صفحات  -

تاریخ انتشار 1993